navigational sign
Sign Language: Towards Sign Understanding for Robot Autonomy
Agrawal, Ayush, Loo, Joel, Zimmerman, Nicky, Hsu, David
Navigational signs are common aids for human wayfinding and scene understanding, but are underutilized by robots. We argue that they benefit robot navigation and scene understanding, by directly encoding privileged information on actions, spatial regions, and relations. Interpreting signs in open-world settings remains a challenge owing to the complexity of scenes and signs, but recent advances in vision-language models (VLMs) make this feasible. To advance progress in this area, we introduce the task of navigational sign understanding which parses locations and associated directions from signs. We offer a benchmark for this task, proposing appropriate evaluation metrics and curating a test set capturing signs with varying complexity and design across diverse public spaces, from hospitals to shopping malls to transport hubs. We also provide a baseline approach using VLMs, and demonstrate their promise on navigational sign understanding. Code and dataset are available on Github.
- Health & Medicine (1.00)
- Education > Curriculum > Subject-Specific Education (0.40)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Robots (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Text Processing (0.47)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.30)
SignLoc: Robust Localization using Navigation Signs and Public Maps
Zimmerman, Nicky, Loo, Joel, Agrawal, Ayush, Hsu, David
To localize, it matches these cues to a large-scale, indoor-outdoor navigation graph, constructed from publicly available maps. Abstract -- Navigation signs and maps, such as floor plans and street maps, are widely available and serve as ubiquitous aids for way-finding in human environments. Y et, they are rarely used by robot systems. This paper presents SignLoc, a global localization method that leverages navigation signs to localize the robot on publicly available maps--specifically floor plans and OpenStreetMap (OSM) graphs-without prior sensor-based mapping. It then employs a probabilistic observation model to match directional and locational cues from the detected signs to the graph, enabling robust topo-semantic localization within a Monte Carlo framework. We evaluated SignLoc in diverse large-scale environments: part of a university campus, a shopping mall, and a hospital complex. Experimental results show that SignLoc reliably localizes the robot after observing only one to two signs. Localizing and navigating in the open world remains a challenge for robots due to the diversity and complexity of human environments.